Universal Deep Neural Network Compression

نویسندگان

  • Yoojin Choi
  • Mostafa El-Khamy
  • Jungwon Lee
چکیده

Compression of deep neural networks (DNNs) for memoryand computation-efficient compact feature representations becomes a critical problem particularly for deployment of DNNs on resource-limited platforms. In this paper, we investigate lossy compression of DNNs by weight quantization and lossless source coding for memory-efficient inference. Whereas the previous work addressed non-universal scalar quantization and entropy coding of DNN weights, we for the first time introduce universal DNN compression by universal vector quantization and universal source coding. In particular, we examine universal randomized lattice quantization of DNNs, which randomizes DNN weights by uniform random dithering before lattice quantization and can perform near-optimally on any source without relying on knowledge of its probability distribution. Entropy coding schemes such as Huffman codes require prior calculation of source statistics, which is computationally consuming. Instead, we propose universal lossless source coding schemes such as variants of Lempel–Ziv–Welch or the Burrows–Wheeler transform. Finally, we present the methods of fine-tuning vector quantized DNNs to recover the performance loss after quantization. Our experimental results show that the proposed universal DNN compression scheme achieves compression ratios of 124.80, 47.10 and 42.46 for LeNet5, 32-layer ResNet and AlexNet, respectively.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Efficient parametrization of multi-domain deep neural networks

A practical limitation of deep neural networks is their high degree of specialization to a single task and visual domain. Recently, inspired by the successes of transfer learning, several authors have proposed to learn instead universal, fixed feature extractors that, used as the first stage of any deep network, work well for several tasks and domains simultaneously. Nevertheless, such universa...

متن کامل

Hfh: Homologically Functional Hashing for Compressing Deep Neural Networks

As the complexity of deep neural networks (DNNs) trends to grow to absorb the increasing sizes of data, memory and energy consumption has been receiving more and more attentions for industrial applications, especially on mobile devices. This paper presents a novel structure based on homologically functional hashing to compress DNNs, shortly named as HFH. For each weight entry in a deep net, HFH...

متن کامل

Cellular Neural Networks for Image Compression by Adaptive Morphological Subband Coding

An “analogic” algorithm is developed for the Cellular Neural Network Universal Machine, that realizes Adaptive Subband Decomposition, a recently proposed algorithm for image compression. In this way, it is possible to obtain very high compression rates with optimized perceptual performance, using real-time mixed analog/digital programmable processors.

متن کامل

Towards Image Understanding from Deep Compression without Decoding

Motivated by recent work on deep neural network (DNN)-based image compression methods showing potential improvements in image quality, savings in storage, and bandwidth reduction, we propose to perform image understanding tasks such as classification and segmentation directly on the compression representations produced by these compression methods. Since the encoders and decoders in DNN-based c...

متن کامل

INFLUENCE OF FIBER ASPECT RATIO ON SHEAR CAPACITY OF DEEP BEAMS USING ARTIFICIAL NEURAL NETWORK TECHNIQUE

This paper deals with the effect of fiber aspect ratio of steel fibers on shear strength of steel fiber reinforced concrete deep beams loaded with shear span to depth ratio less than two using the artificial neural network technique. The network model predicts reasonably good results when compared with the equation proposed by previous researchers. The parametric study invol...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • CoRR

دوره abs/1802.02271  شماره 

صفحات  -

تاریخ انتشار 2018